Skip to main content

Vector Space


Defining the reach of a line

In a geometric context, a line LL in a Euclidean space (e.g., R2\mathbb{R}^2 or R3\mathbb{R}^3) is typically defined as the set of all points that can be obtained by starting from a point p\mathbf{p} on the line and moving along a particular direction vector d\mathbf{d}. This is expressed parametrically as follows:

L={p+td:tR}L = \{ \mathbf{p} + t\mathbf{d} : t \in \mathbb{R} \}

Here:

  • p\mathbf{p} is a vector representing a specific point on the line.
  • d\mathbf{d} is a non-zero vector indicating the direction of the line.
  • tt is a scalar parameter that varies over all real numbers R\mathbb{R}.

As tt varies, p+td\mathbf{p} + t\mathbf{d} generates all the points on the line LL. When t=0t = 0, the expression equals p\mathbf{p}, so the line passes through point p\mathbf{p}. As tt increases or decreases, we move along the line in the direction of d\mathbf{d} or opposite to d\mathbf{d}, respectively.

The eight properties of vector spaces can be organized under the following headings:

Defining the set of vectors orthogonal to a given line

An orthogonal vector to the line LL is a vector that is perpendicular to any vector lying on LL, or more specifically, perpendicular to the direction vector d\mathbf{d} of the line LL.

The set of all vectors orthogonal to the line LL defined by the direction vector d\mathbf{d} is given by the set:

O={vRn:vd=0}O = \{ \mathbf{v} \in \mathbb{R}^n : \mathbf{v} \cdot \mathbf{d} = 0 \}

Here:

  • v\mathbf{v} is a vector in the n-dimensional Euclidean space Rn\mathbb{R}^n.
  • \cdot denotes the dot product.
  • d\mathbf{d} is the direction vector of the line LL.

In this set definition, vd=0\mathbf{v} \cdot \mathbf{d} = 0 is the condition for orthogonality between the vector v\mathbf{v} and the direction vector d\mathbf{d} of the line LL.


Defining the line in a Euclidean space that does not pass through the origin

In Euclidean space, a line in a plane that does not pass through the origin can be represented parametrically or in a normal form.

  1. Parametric form: A line in a plane can be represented parametrically as:

L={p+td:tR}L = \{ \mathbf{p} + t\mathbf{d} : t \in \mathbb{R} \}

where:

  • p\mathbf{p} is a vector representing a specific point on the line,
  • d\mathbf{d} is a non-zero vector indicating the direction of the line,
  • tt is a scalar parameter that varies over all real numbers R\mathbb{R}.

In this form, if p\mathbf{p} is not the zero vector and d\mathbf{d} is not parallel to p\mathbf{p}, the line will not pass through the origin.

  1. Normal form: In the normal form, LL represents a line in the plane, and is defined as the set of all vectors v\mathbf{v} in R2\mathbb{R}^2 that satisfy a particular linear equation. The equation n1x+n2y=dn_1x + n_2y = d is the normal form of the line, where (n1,n2)(n_1, n_2) is the normal vector to the line, and dd is the constant term.

Here's the formal definition of a line LL in a plane that does not pass through the origin, using the given form:

L={v=(x,y)R2n1x+n2y=d,d0}L = \{ \mathbf{v} = (x, y) \in \mathbb{R}^2 \, | \, n_1x + n_2y = d, \, d \neq 0 \}

In this definition:

  • v=(x,y)\mathbf{v} = (x, y) represents a point on the line.
  • (n1,n2)(n_1, n_2) is the normal vector to the line.
  • dd is a nonzero constant term, ensuring that the line does not pass through the origin.

The condition d0d \neq 0 ensures that the line does not pass through the origin, as when d=0d = 0, the line would pass through the origin.

Each of these forms gives a way to represent a line in a plane that does not pass through the origin, using different types of information about the line.


The properties of a vector space

  1. Abelian Groups (related to vector addition):

    • Closure under Addition: u+vV\mathbf{u} + \mathbf{v} \in V
    • Associativity of Addition: (u+v)+w=u+(v+w)(\mathbf{u} + \mathbf{v}) + \mathbf{w} = \mathbf{u} + (\mathbf{v} + \mathbf{w})
    • Commutativity of Addition: u+v=v+u\mathbf{u} + \mathbf{v} = \mathbf{v} + \mathbf{u}
    • Existence of Additive Identity: There exists a vector 0\mathbf{0} in VV such that for all vectors u\mathbf{u} in VV, u+0=u\mathbf{u} + \mathbf{0} = \mathbf{u}.
    • Existence of Additive Inverse: For every vector u\mathbf{u} in VV, there exists a vector u-\mathbf{u} in VV such that u+(u)=0\mathbf{u} + (-\mathbf{u}) = \mathbf{0}.
  2. Scalar Multiplication:

    • Closure under Scalar Multiplication: cuVc\mathbf{u} \in V
  3. Distributive Laws:

    • Distributivity of Scalar Multiplication with respect to Vector Addition: c(u+v)=cu+cvc(\mathbf{u} + \mathbf{v}) = c\mathbf{u} + c\mathbf{v}
    • Distributivity of Scalar Multiplication with respect to Scalar Addition: (a+b)u=au+bu(a + b)\mathbf{u} = a\mathbf{u} + b\mathbf{u}

These properties together provide a structure that ensures a set of vectors, along with the operations of vector addition and scalar multiplication, forms a vector space.

However, apart from proving that a given space is a vector space and the task of proving these axioms it doesn't seem helpful to memorise these axioms. Instead, I think a better approach is to memorise the three larger subdivisions (vector addition, multiplication, and the distributive laws).


Subspace

Sometimes we are presented with a particular equation and we need to know if we can do 'vector' stuff to it (i.e. does this equation represent a vector space). We can do this by working through the characterization of subspaces within a vector space.

  1. Zero Vector Inclusion: The zero vector 0\mathbf{0} is an element of UU, i.e., UU contains the zero vector from the vector space Rn\mathbb{R}^n. 0U\mathbf{0} \in U

  2. Closure under Scalar Multiplication: For any vector u\mathbf{u} in UU and any scalar λ\lambda from R\mathbb{R}, the product λu\lambda\mathbf{u} is also an element of UU. uU,λR    λuU\mathbf{u} \in U, \, \lambda \in \mathbb{R} \implies \lambda\mathbf{u} \in U

  3. Closure under Vector Addition: For any two vectors u,v\mathbf{u}, \mathbf{v} in UU, their sum u+v\mathbf{u} + \mathbf{v} is also an element of UU. u,vU    u+vU\mathbf{u}, \mathbf{v} \in U \implies \mathbf{u} + \mathbf{v} \in U

These characterizations define a subspace UU within a vector space Rn\mathbb{R}^n. They ensure that UU itself forms a vector space under the same vector addition and scalar multiplication operations as Rn\mathbb{R}^n. By satisfying these properties, UU exhibits the essential structure of a vector space, making it a subspace of Rn\mathbb{R}^n.

Worked through example

Let's consider a subset UU of R3\mathbb{R}^3 defined by all vectors that satisfy a particular linear equation, say 2xy+z=02x - y + z = 0. The subset UU can be described as:

U={(x,y,z)R32xy+z=0}U = \{ (x, y, z) \in \mathbb{R}^3 \, | \, 2x - y + z = 0 \}

Now, let's verify if UU is a subspace of R3\mathbb{R}^3 by checking the properties:

  1. Zero Vector Inclusion: The vector (0,0,0)(0, 0, 0) satisfies the equation 2xy+z=02x - y + z = 0, hence (0,0,0)U(0, 0, 0) \in U.

  2. Closure under Scalar Multiplication: Let's take a vector u=(x,y,z)\mathbf{u} = (x, y, z) from UU which satisfies 2xy+z=02x - y + z = 0, and a scalar λ\lambda from R\mathbb{R}. The vector λu=(λx,λy,λz)\lambda\mathbf{u} = (\lambda x, \lambda y, \lambda z) should also satisfy the equation: 2(λx)(λy)+(λz)=λ(2xy+z)=02(\lambda x) - (\lambda y) + (\lambda z) = \lambda(2x - y + z) = 0 Since u\mathbf{u} is in UU, 2xy+z=02x - y + z = 0, and thus λu\lambda\mathbf{u} is also in UU.

  3. Closure under Vector Addition: Let's take two vectors u=(x1,y1,z1)\mathbf{u} = (x_1, y_1, z_1) and v=(x2,y2,z2)\mathbf{v} = (x_2, y_2, z_2) from UU. Their sum u+v=(x1+x2,y1+y2,z1+z2)\mathbf{u} + \mathbf{v} = (x_1 + x_2, y_1 + y_2, z_1 + z_2) should also satisfy the equation: 2(x1+x2)(y1+y2)+(z1+z2)=(2x1y1+z1)+(2x2y2+z2)=0+0=02(x_1 + x_2) - (y_1 + y_2) + (z_1 + z_2) = (2x_1 - y_1 + z_1) + (2x_2 - y_2 + z_2) = 0 + 0 = 0 Since u\mathbf{u} and v\mathbf{v} are in UU, 2x1y1+z1=02x_1 - y_1 + z_1 = 0 and 2x2y2+z2=02x_2 - y_2 + z_2 = 0, and thus u+v\mathbf{u} + \mathbf{v} is also in UU.

Since all three properties are satisfied, UU is indeed a linear subspace of R3\mathbb{R}^3.

Example of a non-linear subspace

A subset defined by an equation involving a non-linear term like (x^2) is not a linear subspace. In vector spaces and their subspaces, the operations of vector addition and scalar multiplication must always yield elements within the set, and this requires the set to be defined linearly.

Let's consider a set (U) in (\mathbb{R}^2) defined by the equation (x^2 + y = 0):

U={(x,y)R2x2+y=0}U = \{(x, y) \in \mathbb{R}^2 \, | \, x^2 + y = 0\}

Now, let's check the properties of subspaces:

  1. Zero Vector Inclusion: The vector (0,0) is in (U) as (0^2 + 0 = 0).

  2. Closure under Scalar Multiplication: Let's take a vector ((x, y)) from (U) and a scalar (\lambda). The vector (\lambda(x, y) = (\lambda x, \lambda y)) results in the equation ((\lambda x)^2 + \lambda y = \lambda^2 x^2 + \lambda y \neq 0) unless (\lambda = 0). This violates the property of closure under scalar multiplication.

  3. Closure under Vector Addition: Let's consider two vectors ((x_1, y_1)) and ((x_2, y_2)) from (U). Their sum is ((x_1 + x_2, y_1 + y_2)). Substituting into the defining equation of (U), we get ((x_1 + x_2)^2 + (y_1 + y_2) = x_1^2 + 2x_1x_2 + x_2^2 + y_1 + y_2). This does not satisfy the equation (x^2 + y = 0) unless (2x_1x_2 + y_1 + y_2 = 0), which is not guaranteed for all (x_1, x_2, y_1, y_2 \in \mathbb{R}). So this violates the property of closure under vector addition.

Due to these violations, the set (U) defined by the equation (x^2 + y = 0) is not a linear subspace of (\mathbb{R}^2).


Span

The span of a set of vectors is the set of all linear combinations of the vectors in the set. It's a fundamental concept in linear algebra, capturing the idea of generating a space from a set of vectors. Formally, the span of a set of vectors {v1,v2,,vk}\{ \mathbf{v}_1, \mathbf{v}_2, \ldots, \mathbf{v}_k \} is defined as follows:

span(v1,v2,,vk)={c1v1+c2v2++ckvkc1,c2,,ckR}\text{span}(\mathbf{v}_1, \mathbf{v}_2, \ldots, \mathbf{v}_k) = \left\{ c_1\mathbf{v}_1 + c_2\mathbf{v}_2 + \ldots + c_k\mathbf{v}_k \, | \, c_1, c_2, \ldots, c_k \in \mathbb{R} \right\}

Here:

  • v1,v2,,vk\mathbf{v}_1, \mathbf{v}_2, \ldots, \mathbf{v}_k are the vectors in the set,
  • c1,c2,,ckc_1, c_2, \ldots, c_k are scalars from the real numbers R\mathbb{R},
  • The expression c1v1+c2v2++ckvkc_1\mathbf{v}_1 + c_2\mathbf{v}_2 + \ldots + c_k\mathbf{v}_k represents a linear combination of the vectors in the set.

The span of these vectors comprises all possible linear combinations of the vectors, and represents a subspace of the vector space that the vectors belong to.


Inner Product (or dot product)

The inner product is a binary operation on a vector space that generalizes the dot product. In a real vector space, an inner product is a function that takes two vectors u\mathbf{u} and v\mathbf{v} and returns a real number, denoted u,v\langle \mathbf{u}, \mathbf{v} \rangle, satisfying the following properties:

  1. Positive Definiteness: For any vector u\mathbf{u} in the vector space, the inner product of u\mathbf{u} with itself is non-negative, and it is zero if and only if u\mathbf{u} is the zero vector: u,u0\langle \mathbf{u}, \mathbf{u} \rangle \geq 0 u,u=0 if and only if u=0\langle \mathbf{u}, \mathbf{u} \rangle = 0 \, \text{ if and only if } \, \mathbf{u} = \mathbf{0}

  2. Symmetry: The inner product is commutative, meaning that swapping the order of the vectors does not change the result: u,v=v,u\langle \mathbf{u}, \mathbf{v} \rangle = \langle \mathbf{v}, \mathbf{u} \rangle

  3. Linearity: The inner product is linear in each argument. This property is often split into two separate properties:

    • Homogeneity in each argument: cu,v=cu,v\langle c\mathbf{u}, \mathbf{v} \rangle = c \langle \mathbf{u}, \mathbf{v} \rangleu,cv=cu,v\langle \mathbf{u}, c\mathbf{v} \rangle = c \langle \mathbf{u}, \mathbf{v} \rangle for all scalars cc.
    • Additivity in each argument: u+w,v=u,v+w,v\langle \mathbf{u} + \mathbf{w}, \mathbf{v} \rangle = \langle \mathbf{u}, \mathbf{v} \rangle + \langle \mathbf{w}, \mathbf{v} \rangleu,v+w=u,v+u,w\langle \mathbf{u}, \mathbf{v} + \mathbf{w} \rangle = \langle \mathbf{u}, \mathbf{v} \rangle + \langle \mathbf{u}, \mathbf{w} \rangle

These properties define a structure on the vector space that allows for the definition of angles and lengths, which are crucial in many areas of mathematics and physics.


Norm's

A norm is a function that assigns a strictly positive length or size to all vectors in a vector space, except for the zero vector, which is assigned a length of zero. It formalizes the notion of length and distance in a vector space. Formally, a norm ||\cdot|| on a vector space VV is a function :VR||\cdot||: V \rightarrow \mathbb{R} satisfying the following properties for all vectors u,v\mathbf{u}, \mathbf{v} in VV and all scalars cc in R\mathbb{R}:

  1. Non-negativity: u0||\mathbf{u}|| \geq 0

  2. Definiteness: u=0 if and only if u=0||\mathbf{u}|| = 0 \, \text{ if and only if } \, \mathbf{u} = \mathbf{0}

  3. Scalar Multiplication: cu=cu||c\mathbf{u}|| = |c| \, ||\mathbf{u}||

  4. Triangle Inequality: u+vu+v||\mathbf{u} + \mathbf{v}|| \leq ||\mathbf{u}|| + ||\mathbf{v}||

Euclidean Norm:

The Euclidean norm (or L2L^2 norm) is a specific type of norm defined on the Euclidean space Rn\mathbb{R}^n. It's derived from the Euclidean inner product, and is defined by the formula:

u=u,u=u12+u22++un2||\mathbf{u}|| = \sqrt{\langle \mathbf{u}, \mathbf{u} \rangle} = \sqrt{u_1^2 + u_2^2 + \ldots + u_n^2}

where u=(u1,u2,,un)\mathbf{u} = (u_1, u_2, \ldots, u_n) is a vector in Rn\mathbb{R}^n.

In simpler terms, the Euclidean norm measures the "ordinary" straight-line distance from the origin to the point identified by the vector u\mathbf{u} in nn-dimensional space. It is a natural extension of the Pythagorean theorem to more than two dimensions. The Euclidean norm is widely used due to its intuitive geometric interpretation and computational simplicity.


Cross-Product of a vector

The cross product is a binary operation on two vectors in three-dimensional space. It produces a third vector that is orthogonal (perpendicular) to the plane of the input vectors. The magnitude of the resulting vector is equal to the area of the parallelogram spanned by the input vectors. The cross product is denoted by u×v\mathbf{u} \times \mathbf{v}, where u\mathbf{u} and v\mathbf{v} are the input vectors.

Formally, if u=(u1,u2,u3)\mathbf{u} = (u_1, u_2, u_3) and v=(v1,v2,v3)\mathbf{v} = (v_1, v_2, v_3), then their cross product is given by:

u×v=(u2v3u3v2,u3v1u1v3,u1v2u2v1)\mathbf{u} \times \mathbf{v} = (u_2v_3 - u_3v_2, u_3v_1 - u_1v_3, u_1v_2 - u_2v_1)

Here's some intuition behind the properties of the cross product:

  1. Orthogonality: The cross product is orthogonal to the original vectors because of the way its components are calculated. Each component of the cross product is computed as the determinant of a 2x2 matrix formed by deleting the row and column of the component's index, which geometrically corresponds to the signed area of the parallelogram formed by the other two components. This ensures that the resulting vector is orthogonal to the plane formed by u\mathbf{u} and v\mathbf{v}.

  2. Magnitude as Area of the Parallelogram: The magnitude of the cross product is given by the formula:

u×v=uvsin(θ)||\mathbf{u} \times \mathbf{v}|| = ||\mathbf{u}|| \, ||\mathbf{v}|| \, \sin(\theta)

where θ\theta is the angle between u\mathbf{u} and v\mathbf{v}. This formula is exactly the formula for the area of a parallelogram, where the sides of the parallelogram are given by the magnitudes of u\mathbf{u} and v\mathbf{v}, and sin(θ)\sin(\theta) gives the "height" of the parallelogram. Hence, the magnitude of the cross product geometrically represents the area of the parallelogram spanned by u\mathbf{u} and v\mathbf{v}.

Why does a cross-product have these properties?

This explanation is illustrating a connection between the cross product, dot product, and determinants in the context of geometric interpretations, particularly relating to the volume of a parallelepiped and the area of a parallelogram.

  1. Defining Vector p\mathbf{p}: The vector p\mathbf{p} is defined such that its dot product with any vector (x,y,z)(x, y, z) is equal to the determinant of the matrix MM with columns formed by vectors (x,y,z)(x, y, z), v=(v1,v2,v3)\mathbf{v} = (v_1, v_2, v_3), and w=(w1,w2,w3)\mathbf{w} = (w_1, w_2, w_3). This determinant is geometrically equal to the volume of the parallelepiped formed by these three vectors.
p(x,y,z)=det(M)\mathbf{p} \cdot (x, y, z) = \det(M)
  1. Volume of Parallelepiped: The volume of a parallelepiped is given by the product of the area of the base and the height. The base is a parallelogram spanned by vectors v\mathbf{v} and w\mathbf{w}, and the height is the component of vector (x,y,z)(x, y, z) that is perpendicular to this parallelogram.

Volume=Area of base×Height\text{Volume} = \text{Area of base} \times \text{Height}

  1. Calculating Height: The height can be found by taking the dot product of (x,y,z)(x, y, z) with a unit vector u\mathbf{u} that is perpendicular to the parallelogram. Thus, the volume of the parallelepiped can also be expressed as:
Volume=Area of base×(u(x,y,z))\text{Volume} = \text{Area of base} \times (\mathbf{u} \cdot (x, y, z))

Comparing this expression to the expression for the determinant of MM, it is evident that:

Area of base×(u(x,y,z))=p(x,y,z)\text{Area of base} \times (\mathbf{u} \cdot (x, y, z)) = \mathbf{p} \cdot (x, y, z)

From this comparison, it's clear that the vector p\mathbf{p} must be equivalent to the product of the area of the base (area of the parallelogram spanned by v\mathbf{v} and w\mathbf{w}) and the unit vector u\mathbf{u} perpendicular to the parallelogram:

Area of base×u=p\text{Area of base} \times \mathbf{u} = \mathbf{p}

This conclusion reveals that vector p\mathbf{p} is:

  1. Perpendicular to both v\mathbf{v} and w\mathbf{w}, as it is aligned with the unit vector u\mathbf{u} which is by definition perpendicular to the parallelogram spanned by v\mathbf{v} and w\mathbf{w}.
  2. Its magnitude is equal to the area of the parallelogram spanned by v\mathbf{v} and w\mathbf{w}.

This geometric interpretation elegantly ties together the algebraic operations of dot product, cross product, and determinants with the geometric concepts of volumes and areas in three-dimensional space.